conv 0
- North America > Canada > Ontario > Toronto (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
Appendices A Dynamic weight sharing
A.1 Noiseless case Each neuron receives the same k-dimensional input x, and its response z This is a strongly convex function, and therefore it has a unique minimum. From Eq. (19) it is clear that w Figure 5: Logarithm of inverse signal-to-noise ratio (mean weight squared over weight variance, see Eq. (6)) for weight sharing objectives in a layer with 100 neurons. B. Dynamics of weight update that uses Eq. (8b) for = 10, different kernel sizes k and. In each iteration, the input is presented for 150 ms. A.2 Biased noiseless case, and its correspondence to the realistic implementation The realistic implementation of dynamic weight sharing with an inhibitory neuron (Section 4.2) introduces a bias in the update rule: Eq. (13) becomes 0 1 X As a result, the final weights are approximately the same among neurons, but have a small norm due to the scaling.
- North America > Canada > Ontario > Toronto (0.04)
- Europe > Italy > Calabria > Catanzaro Province > Catanzaro (0.04)
Appendices A Dynamic weight sharing
A.1 Noiseless case Each neuron receives the same k -dimensional input x, and its response z From Eq. (19) it is clear that Dynamics of weight update that uses Eq. (8b) In each iteration, the input is presented for 150 ms. Realistically, all neurons can't see the same Let's also bound the input mean and noise as E Therefore, we can bound the full gradient by the sum of individual bounds (as it's the Frobenius Both plots in Figure 1 show mean negative log SNR over 10 runs, 100 output neurons each. Learning was performed via SGD with momentum of 0.95. The minimum SNR value was computed from Eq. (5). For our data, the SNR expression in Eq. (6) has The code for both runs is provided in the supplementary material.
Testing Neural Network Verifiers: A Soundness Benchmark with Hidden Counterexamples
Zhou, Xingjian, Xu, Hongji, Xu, Andy, Shi, Zhouxing, Hsieh, Cho-Jui, Zhang, Huan
In recent years, many neural network (NN) verifiers have been developed to formally verify certain properties of neural networks such as robustness. Although many benchmarks have been constructed to evaluate the performance of NN verifiers, they typically lack a ground-truth for hard instances where no current verifier can verify and no counterexample can be found, which makes it difficult to check the soundness of a new verifier if it claims to verify hard instances which no other verifier can do. We propose to develop a soundness benchmark for NN verification. Our benchmark contains instances with deliberately inserted counterexamples while we also try to hide the counterexamples from regular adversarial attacks which can be used for finding counterexamples. We design a training method to produce neural networks with such hidden counterexamples. Our benchmark aims to be used for testing the soundness of NN verifiers and identifying falsely claimed verifiability when it is known that hidden counterexamples exist. We systematically construct our benchmark and generate instances across diverse model architectures, activation functions, input sizes, and perturbation radii. We demonstrate that our benchmark successfully identifies bugs in state-of-the-art NN verifiers, as well as synthetic bugs, providing a crucial step toward enhancing the reliability of testing NN verifiers.
- North America > United States > California > Los Angeles County > Los Angeles (0.14)
- North America > United States > New York > Richmond County > New York City (0.04)
- North America > United States > New York > Queens County > New York City (0.04)
- (6 more...)
- Information Technology > Security & Privacy (0.37)
- Government > Military (0.37)
Prescriptive Process Monitoring in Intelligent Process Automation with Chatbot Orchestration
Zeltyn, Sergey, Shlomov, Segev, Yaeli, Avi, Oved, Alon
Business processes that involve AI-powered automation have been gaining importance and market share in recent years. These business processes combine the characteristics of classical business process management, goal-driven chatbots, conversational recommendation systems, and robotic process automation. In the new context, prescriptive process monitoring demands innovative approaches. Unfortunately, data logs from these new processes are still not available in the public domain. We describe the main challenges in this new domain and introduce a synthesized dataset that is based on an actual use case of intelligent process automation with chatbot orchestration. Using this dataset, we demonstrate crowd-wisdom and goal-driven approaches to prescriptive process monitoring.
- Europe > Austria > Vienna (0.14)
- Asia > Middle East > Israel > Haifa District > Haifa (0.04)
- North America > United States > Hawaii (0.04)
- Information Technology > Artificial Intelligence > Representation & Reasoning > Personal Assistant Systems (1.00)
- Information Technology > Artificial Intelligence > Natural Language > Chatbot (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.47)
Deterministic and Stochastic Analysis of Deep Reinforcement Learning for Low Dimensional Sensing-based Navigation of Mobile Robots
Grando, Ricardo B., de Jesus, Junior C., Kich, Victor A., Kolling, Alisson H., Guerra, Rodrigo S., Drews-Jr, Paulo L. J.
Deterministic and Stochastic techniques in Deep Reinforcement Learning (Deep-RL) have become a promising solution to improve motion control and the decision-making tasks for a wide variety of robots. Previous works showed that these Deep-RL algorithms can be applied to perform mapless navigation of mobile robots in general. However, they tend to use simple sensing strategies since it has been shown that they perform poorly with a high dimensional state spaces, such as the ones yielded from image-based sensing. This paper presents a comparative analysis of two Deep-RL techniques - Deep Deterministic Policy Gradients (DDPG) and Soft Actor-Critic (SAC) - when performing tasks of mapless navigation for mobile robots. We aim to contribute by showing how the neural network architecture influences the learning itself, presenting quantitative results based on the time and distance of navigation of aerial mobile robots for each approach. Overall, our analysis of six distinct architectures highlights that the stochastic approach (SAC) better suits with deeper architectures, while the opposite happens with the deterministic approach (DDPG).
- South America > Brazil (0.04)
- South America > Uruguay (0.04)
A Differentiable Point Process with Its Application to Spiking Neural Networks
This paper is concerned about a learning algorithm for a probabilistic model of spiking neural networks (SNNs). Jimenez Rezende & Gerstner (2014) proposed a stochastic variational inference algorithm to train SNNs with hidden neurons. The algorithm updates the variational distribution using the score function gradient estimator, whose high variance often impedes the whole learning algorithm. This paper presents an alternative gradient estimator for SNNs based on the path-wise gradient estimator. The main technical difficulty is a lack of a general method to differentiate a realization of an arbitrary point process, which is necessary to derive the path-wise gradient estimator. We develop a differentiable point process, which is the technical highlight of this paper, and apply it to derive the path-wise gradient estimator for SNNs. We investigate the effectiveness of our gradient estimator through numerical simulation.
- Asia > Japan > Honshū > Kantō > Tokyo Metropolis Prefecture > Tokyo (0.14)
- North America > United States > New York (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
Certified Defenses: Why Tighter Relaxations May Hurt Training?
Jovanović, Nikola, Balunović, Mislav, Baader, Maximilian, Vechev, Martin
Certified defenses based on convex relaxations are an established technique for training provably robust models. The key component is the choice of relaxation, varying from simple intervals to tight polyhedra. Paradoxically, however, it was empirically observed that training with tighter relaxations can worsen certified robustness. While several methods were designed to partially mitigate this issue, the underlying causes are poorly understood. In this work we investigate the above phenomenon and show that tightness may not be the determining factor for reduced certified robustness. Concretely, we identify two key features of relaxations that impact training dynamics: continuity and sensitivity. We then experimentally demonstrate that these two factors explain the drop in certified robustness when using popular relaxations. Further, we show, for the first time, that it is possible to successfully train with tighter relaxations (i.e., triangle), a result supported by our two properties. Overall, we believe the insights of this work can help drive the systematic discovery of new effective certified defenses.
AgileNet: Lightweight Dictionary-based Few-shot Learning
Ghasemzadeh, Mohammad, Lin, Fang, Rouhani, Bita Darvish, Koushanfar, Farinaz, Huang, Ke
The success of deep learning models is heavily tied to the use of massive amount of labeled data and excessively long training time. With the emergence of intelligent edge applications that use these models, the critical challenge is to obtain the same inference capability on a resource-constrained device while providing adaptability to cope with the dynamic changes in the data. We propose AgileNet, a novel lightweight dictionary-based few-shot learning methodology which provides reduced complexity deep neural network for efficient execution at the edge while enabling low-cost updates to capture the dynamics of the new data. Evaluations of state-of-the-art few-shot learning benchmarks demonstrate the superior accuracy of AgileNet compared to prior arts. Additionally, AgileNet is the first few-shot learning approach that prevents model updates by eliminating the knowledge obtained from the primary training. This property is ensured through the dictionaries learned by our novel end-to-end structured decomposition, which also reduces the memory footprint and computation complexity to match the edge device constraints.
- North America > United States > California > San Diego County > San Diego (0.05)
- Europe > Italy > Veneto > Venice (0.04)